Goto

Collaborating Authors

 Chino Hills


Seamless Interaction: Dyadic Audiovisual Motion Modeling and Large-Scale Dataset

Agrawal, Vasu, Akinyemi, Akinniyi, Alvero, Kathryn, Behrooz, Morteza, Buffalini, Julia, Carlucci, Fabio Maria, Chen, Joy, Chen, Junming, Chen, Zhang, Cheng, Shiyang, Chowdary, Praveen, Chuang, Joe, D'Avirro, Antony, Daly, Jon, Dong, Ning, Duppenthaler, Mark, Gao, Cynthia, Girard, Jeff, Gleize, Martin, Gomez, Sahir, Gong, Hongyu, Govindarajan, Srivathsan, Han, Brandon, He, Sen, Hernandez, Denise, Hristov, Yordan, Huang, Rongjie, Inaguma, Hirofumi, Jain, Somya, Janardhan, Raj, Jia, Qingyao, Klaiber, Christopher, Kovachev, Dejan, Kumar, Moneish, Li, Hang, Li, Yilei, Litvin, Pavel, Liu, Wei, Ma, Guangyao, Ma, Jing, Ma, Martin, Ma, Xutai, Mantovani, Lucas, Miglani, Sagar, Mohan, Sreyas, Morency, Louis-Philippe, Ng, Evonne, Ng, Kam-Woh, Nguyen, Tu Anh, Oberai, Amia, Peloquin, Benjamin, Pino, Juan, Popovic, Jovan, Poursaeed, Omid, Prada, Fabian, Rakotoarison, Alice, Ranjan, Rakesh, Richard, Alexander, Ropers, Christophe, Saleem, Safiyyah, Sharma, Vasu, Shcherbyna, Alex, Shen, Jia, Shen, Jie, Stathopoulos, Anastasis, Sun, Anna, Tomasello, Paden, Tran, Tuan, Turkatenko, Arina, Wan, Bo, Wang, Chao, Wang, Jeff, Williamson, Mary, Wood, Carleigh, Xiang, Tao, Yang, Yilin, Yao, Julien, Zhang, Chen, Zhang, Jiemin, Zhang, Xinyue, Zheng, Jason, Zhyzheria, Pavlo, Zikes, Jan, Zollhoefer, Michael

arXiv.org Artificial Intelligence

Human communication involves a complex interplay of verbal and nonverbal signals, essential for conveying meaning and achieving interpersonal goals. To develop socially intelligent AI technologies, it is crucial to develop models that can both comprehend and generate dyadic behavioral dynamics. To this end, we introduce the Seamless Interaction Dataset, a large-scale collection of over 4,000 hours of face-to-face interaction footage from over 4,000 participants in diverse contexts. This dataset enables the development of AI technologies that understand dyadic embodied dynamics, unlocking breakthroughs in virtual agents, telepresence experiences, and multimodal content analysis tools. We also develop a suite of models that utilize the dataset to generate dyadic motion gestures and facial expressions aligned with human speech. These models can take as input both the speech and visual behavior of their interlocutors. We present a variant with speech from an LLM model and integrations with 2D and 3D rendering methods, bringing us closer to interactive virtual agents. Additionally, we describe controllable variants of our motion models that can adapt emotional responses and expressivity levels, as well as generating more semantically-relevant gestures. Finally, we discuss methods for assessing the quality of these dyadic motion models, which are demonstrating the potential for more intuitive and responsive human-AI interactions.


Business Listing Classification Using Case Based Reasoning and Joint Probability

Sood, Sanjay (AT&T) | Kar, Parijat P. (AT&T)

AAAI Conferences

One challenge of building and maintaining large-scale data management systems is managing data fusion from multiple data sources. Often times, different data sources may represent the same data element in a slightly different way. These differences may represent an error in the data or a disagreement between sources on the correct value that best represents the data point. When the quantity of data managed and fused becomes sufficiently large, manual review becomes impossible, and automated systems must be built to manage data fusion. Some of the traditional solutions use simple voting theory, Dempster-Shafer theory, fuzzy matching and incremental learning. This paper presents a novel approach to data fusion in the domain of business listings. The task at hand, business listing categorization, suffers from conflicting and incomplete data from disparate data sources. Given the need for a high degree of accuracy in this task, we use a combination of case-based reasoning, joint probability, and domain-specific rules to improve data accuracy above other methods.